Learning Structured Outputs via Kernel Dependency Estimation and Stochastic Grammars

نویسندگان

  • Fabrizio Costa
  • Andrea Passerini
  • Paolo Frasconi
چکیده

We focus on graph-valued outputs in supervised learning and propose a novel solution to the pre-image problem in the kernel dependency estimation framework. Output structures are generated by a stochastic grammar and the output feature space is directly associated with the set of productions for the grammar. The regression estimation step learns to map input examples into a feature vector that counts the number of applications of each production rule. A max-propagation algorithm finally builds the predicted output according to the normalized counts. We test our method on a ambiguous context free grammar (CFG) parse tree reconstruction problem. We show on an artificial dataset that mimics the prepositional attachment problem how learning the number of applications of each production rule on a per example base allows CFG parser to better tackle ambiguity issues.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Generalized Kernel Approach to Structured Output Learning

We study the problem of structured output learning from a regression perspective. We first provide a general formulation of the kernel dependency estimation (KDE) approach to this problem using operator-valued kernels. Our formulation overcomes the two main limitations of the original KDE approach, namely the decoupling between outputs in the image space and the inability to use a joint feature...

متن کامل

Machine Translation Using Probabilistic Synchronous Dependency Insertion Grammars

Syntax-based statistical machine translation (MT) aims at applying statistical models to structured data. In this paper, we present a syntax-based statistical machine translation system based on a probabilistic synchronous dependency insertion grammar. Synchronous dependency insertion grammars are a version of synchronous grammars defined on dependency trees. We first introduce our approach to ...

متن کامل

Learning Stochastic Bracketing Inversion Transduction Grammars with a Cubic Time Biparsing Algorithm

We present a biparsing algorithm for Stochastic Bracketing Inversion Transduction Grammars that runs in O(bn3) time instead of O(n6). Transduction grammars learned via an EM estimation procedure based on this biparsing algorithm are evaluated directly on the translation task, by building a phrase-based statistical MT system on top of the alignments dictated by Viterbi parses under the induced b...

متن کامل

Discriminative Machine Learning with Structure

Discriminative Machine Learning with Structure by Simon Lacoste-Julien Doctor of Philosophy in Computer Science and the Designated Emphasis in Communication, Computation and Statistics University of California, Berkeley Professor Michael I. Jordan, Chair Some of the best performing classifiers in modern machine learning have been designed using discriminative learning, as exemplified by Support...

متن کامل

Twenty - Fifth International Conference on Machine Learning

Modeling language at the syntactic or semantic level is a key problem in natural languageprocessing, and involves a challenging set of structured prediction problems. In this talkI’ll describe work on machine learning approaches for syntax and semantics, with a particu-lar focus on lexicalized grammar formalisms such as dependency grammars, tree adjoininggrammars, and catego...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2006